83 research outputs found

    Where do statistical models come from? Revisiting the problem of specification

    Full text link
    R. A. Fisher founded modern statistical inference in 1922 and identified its fundamental problems to be: specification, estimation and distribution. Since then the problem of statistical model specification has received scant attention in the statistics literature. The paper traces the history of statistical model specification, focusing primarily on pioneers like Fisher, Neyman, and more recently Lehmann and Cox, and attempts a synthesis of their views in the context of the Probabilistic Reduction (PR) approach. As argued by Lehmann [11], a major stumbling block for a general approach to statistical model specification has been the delineation of the appropriate role for substantive subject matter information. The PR approach demarcates the interrelated but complemenatry roles of substantive and statistical information summarized ab initio in the form of a structural and a statistical model, respectively. In an attempt to preserve the integrity of both sources of information, as well as to ensure the reliability of their fusing, a purely probabilistic construal of statistical models is advocated. This probabilistic construal is then used to shed light on a number of issues relating to specification, including the role of preliminary data analysis, structural vs. statistical models, model specification vs. model selection, statistical vs. substantive adequacy and model validation.Comment: Published at http://dx.doi.org/10.1214/074921706000000419 in the IMS Lecture Notes--Monograph Series (http://www.imstat.org/publications/lecnotes.htm) by the Institute of Mathematical Statistics (http://www.imstat.org

    Revisiting the Neyman-Scott model: an Inconsistent MLE or an Ill-defined Model?

    Full text link
    The Neyman and Scott (1948) model is widely used to demonstrate a serious weakness of the Maximum Likelihood (ML) method: it can give rise to inconsistent estimators. The primary objective of this paper is to revisit this example with a view to demonstrate that the culprit for the inconsistent estimation is not the ML method but an ill-defined statistical model. It is also shown that a simple recasting of this model renders it well-defined and the ML method gives rise to consistent and asymptotically efficient estimators

    Why the Decision‐Theoretic Perspective Misrepresents Frequentist Inference: Revisiting Stein’s Paradox and Admissibility

    Get PDF
    The primary objective of this paper is to make a case that R.A. Fisher’s objections to the decision‐theoretic framing of frequentist inference are not without merit. It is argued that this framing is congruent with the Bayesian but incongruent with the frequentist approach; it provides the former with a theory of optimal inference but misrepresents the optimality theory of the latter. Decision‐theoretic and Bayesian rules are considered optimal when they minimize the expected loss “for all possible values of θ in Θ” [∀θ∈Θ], irrespective of what the true value θ∗ [state of Nature] happens to be; the value that gave rise to the data. In contrast, the theory of optimal frequentist inference is framed entirely in terms of the capacity of the procedure to pinpoint θ∗. The inappropriateness of the quantifier ∀θ∈Θ calls into question the relevance of admissibility as a minimal property for frequentist estimators. As a result, the pertinence of Stein’s paradox, as it relates to the capacity of frequentist estimators to pinpoint θ∗, needs to be reassessed. The paper also contrasts loss‐based errors with traditional frequentist errors, arguing that the former are attached to θ, but the latter to the inference procedure itself

    THE LINEAR REGRESSION MODEL WITH AUTOCORRELATED ERRORS: JUST SAY NO TO ERROR AUTOCORRELATION

    Get PDF
    This paper focuses on the practice of serial correlation correcting of the Linear Regression Model (LRM) by modeling the error. Simple Monte Carlo experiments are used to demonstrate the following points regarding this practice. First, the common factor restrictions implicitly imposed on the temporal structure of yt and xt appear to be completely unreasonable for any real world application. Second, when one compares the Autocorrelation-Corrected LRM (ACLRM) model estimates with estimates from the (unrestricted) Dynamic Linear Regression Model (DLRM) encompassing the ACLRM there is no significant gain in efficiency! Third, as expected, when the common factor restrictions do not hold the LRM model gives poor estimates of the true parameters and estimation of the ACLRM simply gives rise to different misleading results! On the other hand, estimates from the DLRM and the corresponding VAR model are very reliable. Fourth, the power of the usual Durbin Watson test (DW) of autocorrelation is much higher when the common factor restrictions do hold than when they do not. But, a more general test of autocorrelation is shown to perform almost as well as the DW when the common factor restrictions do hold and significantly better than the DW when the restrictions do not hold. Fifth, we demonstrate how simple it is to, at least, test the common factor restrictions imposed and we illustrate how powerful this test can be.Research Methods/ Statistical Methods,

    Bernoulli Regression Models: Re-examining Statistical Models with Binary Dependent Variables

    Get PDF
    The classical approach for specifying statistical models with binary dependent variables in econometrics using latent variables or threshold models can leave the model misspecified, resulting in biased and inconsistent estimates as well as erroneous inferences. Furthermore, methods for trying to alleviate such problems, such as univariate generalized linear models, have not provided an adequate alternative for ensuring the statistical adequacy of such models. The purpose of this paper is to re-examine the underlying probabilistic foundations of statistical models with binary dependent variables using the probabilistic reduction approach to provide an alternative approach for model specification. This re-examination leads to the development of the Bernoulli Regression Model. Simulated and empirical examples provide evidence that the Bernoulli Regression Model can provide a superior approach for specifying statistically adequate models for dichotomous choice processes.Bernoulli Regression Model, logistic regression, generalized linear models, discrete choice, probabilistic reduction approach, model specification, Research Methods/ Statistical Methods,

    Testing for Structural Breaks and other forms of Non-stationarity: a Misspecification Perspective

    Get PDF
    In the 1980s and 1990s the issue of non-stationarity in economic time series has been in the context of unit roots vs. mean trends in AR(p) models. More recently this perspective has been extended to include structural breaks. In this paper we take a much broader perspective by viewing the problem as one of misspecification testing: assessing the stationarity of the underlying process. The proposed misspecification testing procedure relies on resampling techniques to enhance the informational content of the observed data in an attempt to capture heterogeneity `locally' using rolling window estimators of the primary moments of the stochastic process. The effectiveness of the testing procedure is assessed using extensive Monte Carlo simulationsMaximum Entropy Bootstrap, Non-Stationarity

    What Foundations for Statistical Modeling and Inference?

    Get PDF
    The primary aim of this article is to review the above books in a comparative way from the standpoint of my perspective on empirical modeling and inference. 1 Hacking (1965). Logic of Statistical Inference 2. Mayo (2018). Statistical Inference as Severe Testing: How to Get Beyond the Statistics Wars 3. Conclusion

    Bernoulli Regression Models: Revisiting the Specification of Statistical Models with Binary Dependent Variables

    Get PDF
    The latent variable and generalized linear modelling approaches do not provide a systematic approach for modelling discrete choice observational data. Another alternative, the probabilistic reduction (PR) approach, provides a systematic way to specify such models that can yield reliable statistical and substantive inferences. The purpose of this paper is to re-examine the underlying probabilistic foundations of conditional statistical models with binary dependent variables using the PR approach. This leads to the development of the Bernoulli Regression Model, a family of statistical models, which includes the binary logistic regression model. The paper provides an explicit presentation of probabilistic model assumptions, guidance on model specification and estimation, and empirical application
    • …
    corecore